Incremental Learning Using a Grow-and-Prune Paradigm With Efficient Neural Networks
نویسندگان
چکیده
Deep neural networks (DNNs) have become a widely deployed model for numerous machine learning applications. However, their fixed architecture, substantial training cost, and significant redundancy make it difficult to efficiently update them accommodate previously unseen data. To solve these problems, we propose an incremental framework based on grow-and-prune network synthesis paradigm. When new data arrive, the first grows connections gradients increase capacity Then, iteratively prunes away magnitude of weights enhance compactness, hence recover efficiency. Finally, rests at lightweight DNN that is both ready inference suitable future grow- and-prune updates. The proposed improves accuracy, shrinks size, significantly reduces additional cost incoming compared conventional approaches, such as from scratch fine-tuning. For LeNet-300-100 (LeNet-5) architectures derived MNIST dataset, by up 64 (67), 63 (63), 69 (73 percent) scratch, fine-tuning, respectively. ResNet-18 architecture ImageNet dataset (DeepSpeech2 AN4 dataset), corresponding reductions against fine-tunning, are 60 (62), 72 (71 percent), Our models contain fewer parameters but achieve higher accuracy relative baselines.
منابع مشابه
NeST: A Neural Network Synthesis Tool Based on a Grow-and-Prune Paradigm
Neural networks (NNs) have begun to have a pervasive impact on various applications of machine learning. However, the problem of finding an optimal NN architecture for large applications has remained open for several decades. Conventional approaches search for the optimal NN architecture through extensive trial-and-error. Such a procedure is quite inefficient. In addition, the generated NN arch...
متن کاملrodbar dam slope stability analysis using neural networks
در این تحقیق شبکه عصبی مصنوعی برای پیش بینی مقادیر ضریب اطمینان و فاکتور ایمنی بحرانی سدهای خاکی ناهمگن ضمن در نظر گرفتن تاثیر نیروی اینرسی زلزله ارائه شده است. ورودی های مدل شامل ارتفاع سد و زاویه شیب بالا دست، ضریب زلزله، ارتفاع آب، پارامترهای مقاومتی هسته و پوسته و خروجی های آن شامل ضریب اطمینان می شود. مهمترین پارامتر مورد نظر در تحلیل پایداری شیب، بدست آوردن فاکتور ایمنی است. در این تحقیق ...
Reinforcement Learning in Neural Networks: A Survey
In recent years, researches on reinforcement learning (RL) have focused on bridging the gap between adaptive optimal control and bio-inspired learning techniques. Neural network reinforcement learning (NNRL) is among the most popular algorithms in the RL framework. The advantage of using neural networks enables the RL to search for optimal policies more efficiently in several real-life applicat...
متن کاملReinforcement Learning in Neural Networks: A Survey
In recent years, researches on reinforcement learning (RL) have focused on bridging the gap between adaptive optimal control and bio-inspired learning techniques. Neural network reinforcement learning (NNRL) is among the most popular algorithms in the RL framework. The advantage of using neural networks enables the RL to search for optimal policies more efficiently in several real-life applicat...
متن کاملLearning to Prune Filters in Convolutional Neural Networks
Many state-of-the-art computer vision algorithms use large scale convolutional neural networks (CNNs) as basic building blocks. These CNNs are known for their huge number of parameters, high redundancy in weights, and tremendous computing resource consumptions. This paper presents a learning algorithm to simplify and speed up these CNNs. Specifically, we introduce a “try-and-learn” algorithm to...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Transactions on Emerging Topics in Computing
سال: 2022
ISSN: ['2168-6750', '2376-4562']
DOI: https://doi.org/10.1109/tetc.2020.3037052